Goto

Collaborating Authors

 Bell County


Reinforcement Learning with Dynamic Multi-Reward Weighting for Multi-Style Controllable Generation

de Langis, Karin, Koo, Ryan, Kang, Dongyeop

arXiv.org Artificial Intelligence

Style is an integral component of text that expresses a diverse set of information, including interpersonal dynamics (e.g. formality) and the author's emotions or attitudes (e.g. disgust). Humans often employ multiple styles simultaneously. An open question is how large language models can be explicitly controlled so that they weave together target styles when generating text: for example, to produce text that is both negative and non-toxic. Previous work investigates the controlled generation of a single style, or else controlled generation of a style and other attributes. In this paper, we expand this into controlling multiple styles simultaneously. Specifically, we investigate various formulations of multiple style rewards for a reinforcement learning (RL) approach to controlled multi-style generation. These reward formulations include calibrated outputs from discriminators and dynamic weighting by discriminator gradient magnitudes. We find that dynamic weighting generally outperforms static weighting approaches, and we explore its effectiveness in 2- and 3-style control, even compared to strong baselines like plug-and-play model. All code and data for RL pipelines with multiple style attributes will be publicly available.


Combatting Human Trafficking in the Cyberspace: A Natural Language Processing-Based Methodology to Analyze the Language in Online Advertisements

Perez, Alejandro Rodriguez, Rivas, Pablo

arXiv.org Artificial Intelligence

This project tackles the pressing issue of human trafficking in online C2C marketplaces through advanced Natural Language Processing (NLP) techniques. We introduce a novel methodology for generating pseudo-labeled datasets with minimal supervision, serving as a rich resource for training state-of-the-art NLP models. Focusing on tasks like Human Trafficking Risk Prediction (HTRP) and Organized Activity Detection (OAD), we employ cutting-edge Transformer models for analysis. A key contribution is the implementation of an interpretability framework using Integrated Gradients, providing explainable insights crucial for law enforcement. This work not only fills a critical gap in the literature but also offers a scalable, machine learning-driven approach to combat human exploitation online. It serves as a foundation for future research and practical applications, emphasizing the role of machine learning in addressing complex social issues.


Soros DA put murder case on 'back burner' because it doesn't 'fit' liberal agenda: victim's family

FOX News

Thomas Villarreal of the Austin Police Association discusses the police department's decision to implement artificial intelligence software in an effort to alleviate their officer shortage on "Fox & Friends Weekend." The family of a man killed in one of Austin, Texas' most infamous shootings blasted the local district attorney for putting the case on the "back burner" because it didn't fit his progressive agenda. Travis County District Attorney Jose Garza, funded by left-wing billionaire George Soros, is letting the nearly two-year case languish and is instead prioritizing cases that fit a political agenda, said Nick Kantor, whose brother, Doug, was killed in gang crossfire on June 12, 2021, that left more than a dozen innocent bystanders wounded. Doug Kantor, then 25 and working for Ford Motor Co., was visiting Austin from Michigan to celebrate earning his master's degree with friends when two rival gangs of teenagers from Killeen, Texas, opened fire on each other in the city's packed Sixth Street entertainment and nightlife hub. Doug Kantor, a New York native who had just bought a new home and was set to marry his high school sweetheart, was killed in the shooting and 13 other innocent bystanders were injured in the hail of bullets from both gangs that became the largest mass casualty incident in Austin in about a decade.


2022 military hardware to remember

FOX News

Rep. Rob Wittman, R-Va., joins'Fox News Live' to react to the United States Air Force's unveiling of its new B-21 raider stealth bomber, named'The Raider' for Jimmy Doolittle's famous bombing raid on Japan in WW2. With the launch of the Air Force's hypersonic missile off the coast of California earlier this month, the Navy's development of water-based drones over the summer and the recent unveiling of the B-21 Raiders, the U.S. military has made major technological advancements over the past year. The military unveiled the U.S. Air Force B-21 Raider in Palmdale, California. The B-21 Raider is the first new American bomber aircraft in more than three decades. In an email to Fox News Digital, a spokesperson confirmed the Air Force would transition its three-bomber fleet to a two-bomber fleet of B-21s and modernized B-52s.


IELM: An Open Information Extraction Benchmark for Pre-Trained Language Models

Wang, Chenguang, Liu, Xiao, Song, Dawn

arXiv.org Artificial Intelligence

We introduce a new open information extraction (OIE) benchmark for pre-trained language models (LM). Recent studies have demonstrated that pre-trained LMs, such as BERT and GPT, may store linguistic and relational knowledge. In particular, LMs are able to answer ``fill-in-the-blank'' questions when given a pre-defined relation category. Instead of focusing on pre-defined relations, we create an OIE benchmark aiming to fully examine the open relational information present in the pre-trained LMs. We accomplish this by turning pre-trained LMs into zero-shot OIE systems. Surprisingly, pre-trained LMs are able to obtain competitive performance on both standard OIE datasets (CaRB and Re-OIE2016) and two new large-scale factual OIE datasets (TAC KBP-OIE and Wikidata-OIE) that we establish via distant supervision. For instance, the zero-shot pre-trained LMs outperform the F1 score of the state-of-the-art supervised OIE methods on our factual OIE datasets without needing to use any training sets. Our code and datasets are available at https://github.com/cgraywang/IELM


US Army fires Javelin anti-tank missiles from robots in key tech test

FOX News

Fox News Flash top headlines are here. Check out what's clicking on Foxnews.com. The U.S. Army test-fired Javelin anti-tank missiles at a recent exhibition in Fort Hood, Texas to demonstrate technological advancement in its fighting capabilities. During a series of weapons drills and exercises, soldiers fired Javelins and .50-caliber A Javelin missile fired by soldiers with the 2nd Stryker Brigade Combat Team, separate from the exhibition in Texas.


Future robot battle buddies may read your emotions to fight better

#artificialintelligence

The Army's plans for robotic wingmen in vehicle formations, a drone on every soldier and robotic mules carrying gear all aim to take the load off the fighter. But how will the two communicate, robot and human? Voice commands like automated assistants on smartphones are great, but not when the threat of incoming fire means the robot battle buddy needs to decipher a range of priorities that humans might take for granted. The next test will come in late 2021 and involve a company-sized maneuver at Fort Hood, Texas. Think more C3PO or R2D2 in the "Star Wars" movies than Hal in "2001: A Space Odyssey" --or better yet, a friendly cyborg from "Terminator" might be the best way to see your robot combatant squad mate of the distant future.


Language Models are Open Knowledge Graphs

Wang, Chenguang, Liu, Xiao, Song, Dawn

arXiv.org Artificial Intelligence

This paper shows how to construct knowledge graphs (KGs) from pre-trained language models (e.g., BERT, GPT-2/3), without human supervision. Popular KGs (e.g, Wikidata, NELL) are built in either a supervised or semi-supervised manner, requiring humans to create knowledge. Recent deep language models automatically acquire knowledge from large-scale corpora via pre-training. The stored knowledge has enabled the language models to improve downstream NLP tasks, e.g., answering questions, and writing code and articles. In this paper, we propose an unsupervised method to cast the knowledge contained within language models into KGs. We show that KGs are constructed with a single forward pass of the pre-trained language models (without fine-tuning) over the corpora. We demonstrate the quality of the constructed KGs by comparing to two KGs (Wikidata, TAC KBP) created by humans. Our KGs also provide open factual knowledge that is new in the existing KGs. Our code and KGs will be made publicly available.


Regional Rainfall Prediction Using Support Vector Machine Classification of Large-Scale Precipitation Maps

Hussein, Eslam A., Ghaziasgar, Mehrdad, Thron, Christopher

arXiv.org Machine Learning

Rainfall prediction helps planners anticipate potential social and economic impacts produced by too much or too little rain. This research investigates a class-based approach to rainfall prediction from 1-30 days in advance. The study made regional predictions based on sequences of daily rainfall maps of the continental US, with rainfall quantized at 3 levels: light or no rain; moderate; and heavy rain. Three regions were selected, corresponding to three squares from a $5\times5$ grid covering the map area. Rainfall predictions up to 30 days ahead for these three regions were based on a support vector machine (SVM) applied to consecutive sequences of prior daily rainfall map images. The results show that predictions for corner squares in the grid were less accurate than predictions obtained by a simple untrained classifier. However, SVM predictions for a central region outperformed the other two regions, as well as the untrained classifier. We conclude that there is some evidence that SVMs applied to large-scale precipitation maps can under some conditions give useful information for predicting regional rainfall, but care must be taken to avoid pitfall


Being Progressive Shouldn't Mean Being Anti-Algorithm

#artificialintelligence

Rep. Alexandria Ocasio-Cortez, speaking at an event in January 2019 honoring the legacy of Dr. Luther King, said, "Algorithms are still made by human beings, and those algorithms are still pegged to basic human assumptions. And if you don't fix the bias, then you are just automating the bias." Though her comments were correct--algorithms can indeed reflect and exhibit human bias--Rep. Ocasio-Cortez's framing of the intersection of algorithms and fairness highlighted an often-ignored issue in progressive politics. The political movement, defined in part by its commitment to social justice, is unsurprisingly critical of the potential for algorithms, particularly AI, to facilitate discrimination, yet seemingly pays little attention to the ways in which algorithms can actually reduce discrimination.